Essay/Term paper: Legislative proposal for new indecency language in telecom bill
Essay, term paper, research paper: History Essays
Free essays available online are good but they will not follow the guidelines of your particular writing assignment. If you need a custom term paper on History Essays: Legislative Proposal For New Indecency Language In Telecom Bill, you can hire a professional writer here to write you a high quality authentic essay. While free essays can be traced by Turnitin (plagiarism detection program), our custom written essays will pass any plagiarism test. Our writing service will save you time and grade.
Legislative Proposal for New Indecency Language in Telecom Bill
I. Summary
Although the October 16, 1995 legislative proposal purports to regulate "
computer pornography", the proposal contains fatal flaws which render the
proposal at best counterproductive and at worst devastating to on-line
communications. First, it prohibits, but fails to define, "indecent" speech to
minors -- a dangerously vague, medium-specific, and, after decades of litigation,
still undefined concept, which may include mere profanity. This may tie up
successful prosecution of the law in courts for years to come, while courts
wrestle to divine a constitutional definition of "indecent" -- and while
companies are left with uncertain liability.
Second, the October 16 proposal may actually hold systems liable for
communications over which they have no specific knowledge or control. The
proposal purports to target those who "knowingly" send prohibited communications
-- itself a relatively low standard of liability that may not even require
actual intent or willfulness. Nevertheless, because the proposal i) defines the
elements of criminal liability in vague and contradictory terms, and ii)
eliminates safeharbors in the Senate bill that would define a clear standard of
care, it might hold systems liable for actions that don't reach even a "
knowingly" standard of liability. As a result, access providers, system
managers and operators, and employers may potentially be liable for actions of
users over which they have no specific knowledge, intent, or control.
For any company that communicates by computer, the proposal:
1) Creates liability for, but never defines, "indecent" speech, a dangerously
vague standard that could leave companies criminally liable for use of mere
profanity;
2) Establishes vague and contradictory standards of liability that could leave
innocent companies vicariously liable for communications over which they have no
control;
3) Strips workable affirmative defenses from the Senate bill, eliminating a
clear standard of care for companies.
Not only does the proposal endanger companies, it fails to protect
children. The indecency standard guarantees that enforcement will be tied up in
the courts for years to come. Companies will be particularly reticent to
identify and eradicate prohibited communications when they are incapable of
discerning which communications are "indecent" and when the company's consequent
knowledge of the communications may actually make them liable. At worst, the
proposal will either shut down systems entirely or will shut down any attempts
to constructively monitor and screen systems, as providers take a know-nothing
stance to avoid prosecution for purported knowledge.
II. The "Indecency" Standard and Uncertain and Conflicting Standards of
Culpability Implicate Innocent Companies But Fail To Protect Children.
A. The undefined "indecency" standard is possibly unenforceable and certainly
counterproductive.
Although the October 16 proposal purports to regulate "computer
pornography", it actually prohibits all "indecent" communications by computer or
"telecommunications device" (an undefined term that presumably includes
telephones and facsimiles) to persons under 18. Because the term "indecent" is
a medium-specific term that, after decades of litigation, remains undefined, it
is uncertain precisely what would be prohibited by this section. In the context
of broadcasting, the Supreme Court has defined mere expletives as indecent See
FCC v. Pacifica Foundation, 438 U.S. 726 (1978).: Would the use of an expletive
in a communication that is made available to a minor trigger a criminal felony?
An illustration. After this law passes, a 17-year old college freshman
is writing a paper on the "indecency". He decides to look at Supreme Court cases
to determine what he is prohibited from seeing. The university librarian, who
believes the student looks young for a freshman, directs the student to the
Supreme Court Pacifica case, which defined "indecency" for the purpose of
broadcast media. If the librarian directs the student to the bound version of
the Supreme Court Reporter, she has done her job well. If she sends an
electronic version on-line, she goes to federal prison for 5 years. The
Pacifica case contains as an appendix a transcript of the George Carlin
monologue on "Seven Dirty Words", which the Court found indecent for purposes of
broadcasting.
The Supreme Court had no qualms about printing the case, because it was
in a different medium than broadcasting -- one requiring someone to access it
and requiring literacy. The October 16 proposal recognizes no such distinction
between media, however. Nor does it define "indecency". Indeed, it treats all "
indecency" as "pornography". Would the Pacifica case be banned from on-line
access by our schools and libraries by the October 16 proposal? It would by any
normally prudent access provider who wanted to avoid the possibility of spending
5 years in federal prison.
Other examples: (i) a sender posts a message to a Bulletin Board that
contains an expletive or a medical or literary passage that is "indecent" and is
then read by a minor; (ii) a university provides on-line access to all students,
including some freshmen under the age of 18, to its library, including works
containing "indecent" passages; (iii) a company that employs a high school
senior as an intern knowingly posts a message from an employee that contains
some of the "Seven Dirty Words" on an employee bulletin board. Under a plain
language reading of the proposal, any of these actions might subject the sender
to a criminal felony conviction. Given such potential liability, companies may
be faced with avoiding liability by either shutting down screening of
communications, or shutting down systems entirely.
At best, the indecency provisions are simply unenforceable. In
regulating indecent speech, the courts have held that the government must take
into account the medium being regulated, must use the least restrictive means to
further its articulated interest, and may not curtail all adult discourse to
only what is fit for children. Sable Communications of California, Inc. v. FCC,
492 U.S. 115, 126, 128 (1989). The Department of Justice noted that the
language upon which Sec. (d) of the proposal is based raises constitutional
questions due to the lack of criminal intent required for the age element.
Letter from Kent Markus, Acting Assistant Attorney General, to Sen. Leahy (June
13, 1995), 141 Cong. Rec. S 8344. The Justice Department stated its concern
that "this subsection would consequently have the effect of regulating indecent
speech between consenting adults". Such a holding by a court could render the
indecency standard constitutionally unenforceable.
The indecency standard is counterproductive. First, it ensures that
rather than effectively protecting children on the Internet, the law will be
caught up in fruitless litigation for years to come. The much less expansive
statutory limitations and subsequent FCC regulations on dial-a-porn engendered
ten years of litigation before a constitutional standard was established.
Second, companies are apt in the face of uncertain liability and an
undefined standard of "indecency" to abdicate any positive role in screening
rather than risk liability for discovered or imputed knowledge. Companies would
be particularly vulnerable during the years of litigation it would take to
establish a constitutional standard of "indecency" by computer communications.
At worst, the indecency provisions would shut down entire networks.
At the very least, the indecency standard establishes a separate
standard of liability for the Net, relegating it to second class citizenship
among all media. Information which is freely available in bookstores, libraries,
and record shops could be banned on the Internet. The electronic editions of
newspapers could at times be prevented from publishing stories appearing in the
printed version.
In place of a nebulous indecency standard, children would be far better
protected by a "harmful to minors" standard that spells out explicitly what type
of material is prohibited. Such a standard is currently in place in all 50
states and in the District of Columbia and has been upheld consistently be the
courts.
B. Vague and contradictory standards of liability threaten innocent companies.
The dangerously vague "indecency" standard is compounded by vague and
contradictory criminal elements in the Title 18 and Title 47 offenses.
According to a former federal prosecutor in our firm, depending upon how courts
read such ambiguous elements, innocent companies might be left vicariously
liable for communications over which they have no specific knowledge or control.
This danger is particularly acute given the incredibly large amount of
information that flows over systems and the utter impossibility of companies to
screen, review, and remove all "indecent" communications -- even if they could
define such communications. Imagery and graphics are particularly troublesome,
as they can be screened only by the old fashioned way -- by human inspection,
conceivably necessitating an indecency inspector at every company using on-line
systems. 1. Vague and Contradictory Standards of Intent and Control Subsection
(d)(1) holds a person or company liable for "knowingly making available" any
prohibited communication, "regardless of whether the maker of such communication
placed the call or initiated the communication(s)". Disturbingly, "knowingly"
and "makes available" are undefined. According to a former federal prosecutor at
our firm, "knowingly" is a relatively low standard of liability, that does not
require willfulness or intent.
The standard of duty to prevent communications once a company is on
notice that they exist is unclear. If notified that a potentially offending
communication exists on a bulletin board on the system, is the system manager
now culpable of "knowingly . . . making vailable" the communication? If
notified that an offending communication exists somewhere on a company's system,
is there then a duty to hunt for the material and delete it? Once given notice,
is there a duty to prevent retransmission? These problems are compounded
because even if a company is informed of the existence of an offending
communication, it may not know whether the communication is "indecent". Indeed,
the company may be precluded by state, local, or federal privacy statutes or
other laws from interfering with or even reviewing the communication.
The Title 18 offense and the Sec. (d) offense lack crucial elements
provided in the Sec. (a) offense that are necessary to ensure that companies are
held liable only for communications that they exert control over and intend to
send. Specifically, Sec. (a) provides that a sender must knowingly both (i) "
make[], create[], solicit[]" and (ii) "purposefully make[] available" or "
initiate[] the transmission of" a communication in order to be held liable for
it. Courts would presumably attempt to reconcile the differences in identical
crimes in the same bill in a way that gives meaning to each word of the
legislation. Consequently, courts may read the lack of such elements in the
Title 18 and Sec. (d) offenses to implicate company-operated systems by
vicarious liability for the actions of users.
2. Vague and Contradictory Standards of Knowledge.
Furthermore, the Title 18 and Title 47 indecency to minors provisions
create vague and inexplicably conflicting standards of culpability as to the age
of a communication recipient. Both sections begin with a "knowingly"
requirement. The Title 18 provision, however, requires in addition that the
communicator or transmitter "believes" that the recipient has not attained the
age of 18, and "know(s)" that the communication "will be obtained by a person
believed to be under 18 years of age". The Title 47 provision contains no such
additional requirements.
The Title 18 offense itself is dangerously vague on whether specific or
general knowledge of the recipient is required. If a communication is posted to
a bulletin board to which the sender "believes" or "knows" that children have
access, is the sender in violation? Is the bulletin board operator? Is the
system upon which the bulletin board is located?
Even more disturbing is the discrepancy between the elements of
liability in Titles 18 and 47. Again, courts would presumably attempt to
reconcile discrepancies in identical crimes in the same bill in a way that gives
meaning to each word of the legislation. Consequently, courts may read the
statute to establish that the level of knowledge or belief required to establish
liability under the Title 18 provision is greater than the level required for
liability under the Title 47 provision. Thus, someone might be prosecuted under
Title 47 despite the fact that he does not believe the recipient of a
communication is a minor, and despite the fact that he does not know whether the
communication will actually be received by a minor. Such a reading would be
supported by the fact that the Title 18 offense is punishable by a longer term
(5 years) than the Title 47 offense (2 years).
This standard is particularly troublesome for companies that operate
systems or bulletin boards that have the capacity of being accessed by minors,
as do nearly all systems or bulletin boards interconnected by the Internet. If
one need not know whether the recipient of a communication is a minor, or
whether a communication will actually be received by a minor, posting a
communication to a system potentially accessible by a minor, which in fact is
accessed by a minor, may render one liable, under such a reading, under the
Title 47 offense.
C. Sec. (d)(2) Protections for Companies Gutted.
As drafted, Sec. (d)(1) effectively guts the protections that Sec.
(d)(2) is intended to provide to businesses and other systems. Sec. (d)(2)
establishes protection against vicarious liability for system operators and
managers under Sec. (d)(1), by limiting liability for a "telecommunications
facilities" under one's control to where one has "knowingly permit(ted)" the
facility to be used for a prohibited Sec. (d)(1) purpose, "with the intent" that
it be so used. Sec. 223(d)(2). This protection is particularly important given
the recent court holding in Stratton Oakmont that systems may be liable for
every single communication sent over their network, regardless of their
knowledge of the nature of the communication. Stratton Oakmont Inc. v. Prodigy
Services Co., No. E31063/94 (N.Y. Sup. Ct. May 24, 1995).
The offense in Sec. (d)(1) is so broadly drawn, however, that it guts
this defense. Sec. (d)(1) holds liable anyone who "makes or makes available" a
prohibited communication, "regardless of whether the maker of such communication
placed the call or initiated the communication". Sec. 223(d)(1). Any Sec.
(d)(2) offense would presumably entail a violation of this provision. Thus,
rather than being protected by a higher standard of liability, facilities could
be doubly liable, under Sec.s (d)(1) and (d)(2), for a prohibited message sent
by a user.
D. Affirmative Defenses Gutted.
Although the October 16 proposal's authors purport to hold liable only
systems or access providers that knowingly transmit prohibited communications --
itself a low threshold -- the proposal guts safeguards in the Senate-passed
telecommunications bill that would have ensured even that:
1. Mere Provision of Access.
First, the proposal strips a Senate defense that would protect access
providers against liability "solely for providing access" to a network or system
not under their control. (Subsec. 402(f)(1).) Given the uncertainties of
application of the "knowingly" standard, this defense is necessary to ensure
that access providers are not held liable for material of which they have no
knowledge or over which they have no ontrol.
2. Employer Defense.
Second, the proposal strips a Senate defense that would protect
employers from being held liable for the unauthorized actions of a rogue
employee. The Senate-passed bill established that employers shall not be held
liable for the actions of an employee or an agent such as a subcontractor unless
the employee or agent's conduct is "within the scope of his employment or agency
and the employer has knowledge of, authorizes, or ratifies the employees or
agent's conduct". (Subsec. 402(f)(2)). A former federal prosecutor in our firm
indicates that absent this defense, a company might be held liable under a
theory of agency or vicarious liability for the actions of an employee whether
or not the company intended those actions.
3. Screening and Compliance With FCC Regulations.
The sole remaining affirmative defense, which provides protection from
prosecution under Sec. (d) for compliance with access restrictions and
subsequent FCC regulations, is worthless to companies. First, this defense is
meaningless without a comparable defense to prosecution under Title 18, for
which companies are liable for even higher penalties (5 years in prison vs. 2
years in prison) for the same behavior (an "indecent" communication to a minor).
The October 16 proposal provides no comparable Title 18 safeharbor, rendering
the Title 47 safeharbor worthless.
Second, the proposal prescribes restrictions with which companies must
comply until FCC regulations take effect, but the restrictions, lifted wholesale
from FCC dial-a-porn regulations, are inapplicable to most companies and would
be impossible to comply with. The interim restrictions require companies to
block or restrict access to any person under 18 through the use of a verified
credit card, adult access code, or adult personal identification number (PIN).
Such restrictions are workable for a dial-a-porn provider who provides
restricted access to a telephone number for a commercial charge. Such
restrictions are antithetical, however, to unrestricted, intentionally open
connections, such as within a company's computer network between systems.
Companies are required to comply with the interim restrictions until FCC
regulations become effective, which, because the proposal restricts
constitutionally protected indecent speech, could take a decade or more. The
dial-a-porn regulations on which the interim restrictions are based took ten
years for constitutionally sustainable regulations to finally take effect. Thus,
companies could be left without a defense for a decade or more, while the FCC
attempts to fashion constitutional regulations -- which may be nevertheless
prove useless to companies. Indeed, if the FCC regulations resemble the interim
restrictions in the proposal, they will in fact be useless to most companies.